"Activists swiftly warned that the policy as it was published would backfire. The policy was vague and had been put together without much input from the communities most vulnerable to harassment and doxxing, the activists argued. They had little faith in Twitter's reporting and appeals process, which they described as unreliable, automated and allowing for little discussion about the enforcement of policies."
"Employees worry that, should Signal fail to build policies and enforcement mechanisms to identify and remove bad actors, the fallout could bring more negative attention to encryption technologies from regulators at a time when their existence is threatened around the world.
"The world needs products like Signal - but they also need Signal to be thoughtful," said Gregg Bernstein, a former user researcher who left the organization this month over his concerns. "It's not only that Signal doesn't have these policies in place. But they've been resistant to even considering what a policy might look like.""
"The development of a new code of conduct will take place in two phases.
The first will include setting policies for in-person and virtual events as well as policies for technical spaces including chat rooms and other Wikimedia projects. It is set to be ratified by the board by 30 August."
"Twitter has said that tweets wishing for Donald Trump's death in the wake of the president's diagnosis with Covid-19 violate its policies and could result in suspension."
"Such bans are possible because school policies allow for the blocking of traffic to certain websites on campus wifi networks, measures that are typically reserved for harmful content and pornography. But those policies can also extend to specific apps, which has been done in the past with platforms like anonymous social media account Yik Yak."
"During the COVID-19 pandemic, millions of Americans will be forced to work, play, and learn from home for the foreseeable future. Such a massive shift will lean not only on shaky and expensive U.S. broadband networks, but popular teleconferencing programs that often don't quite work as advertised.
Zoom in particular has seen a flood of new users, and the company's stock has jumped roughly 20 percent since the COVID-19 outbreak began. But as new users flock to the platform for work, they should be aware of a few things: namely, the company's data collection, its shaky privacy policy, and the fact your boss knows when you're not giving them your undivided attention."
"Journal editors, researchers and publishers are now debating the place of such AI tools in the published literature, and whether it's appropriate to cite the bot as an author. Publishers are racing to create policies for the chatbot, which was released as a free-to-use tool in November by tech company OpenAI in San Francisco, California."
"The new policy is a major shift in Twitter's efforts to balance its ideological commitment to free expression with user demands for improved enforcement of rules against harassment, hate speech and other toxic behavior."
""If we build highly advanced autonomous AI, we risk creating systems that autonomously pursue undesirable goals", adding that "we may not be able to keep them in check".
Other policy recommendations in the document include: mandatory reporting of incidents where models show alarming behaviour; putting in place measures to stop dangerous models from replicating themselves; and giving regulators the power to pause development of AI models showing dangerous behaviours."
""We believe the government should introduce legislation that ensures commercial images are labelled with a logo where any part of the body, including its proportions and skin tone, are digitally altered," its report says.
Meanwhile, it says dermal fillers should be made prescription-only substances, in line with Botox, and there should be minimum training standards for providers."
"Critics of the government's flagship internet regulation policy are warning it could lead to a North Korean-style censorship regime, where regulators decide which websites Britons are allowed to visit, because of how broad the proposals are."
"Service providers, including many ORG members, will be required to do this through the imposition of a "duty of care" - a concept awkwardly borrowed from health & safety - which will require them to monitor the integrity of their services not by objective technical standards, but by subjective "codes of practice" on both illegal and legal content.
Although the framework has been drawn up with large American social media platforms in mind, it would apply to any site or service with UK users which hosts user-generated content. A blog with comments will be fair game. An app with user reviews will be fair game. "
"Facebook has lifted a ban on posts claiming Covid-19 was man-made, following a resurgence of interest in the "lab leak" theory of the disease's onset.
The social network says its new policy comes "in light of ongoing investigations into the origin"."
""The fact is that an interconnected ecosystem of companies and data brokers, of purveyors of fake news and peddlers of division, of trackers and hucksters just looking to make a quick buck, is more present in our lives than it has ever been," he said. "Technology does not need vast troves of personal data, stitched together across dozens of websites and apps, in order to succeed.""
"Computers once performed facial recognition rather imprecisely, by identifying people's facial features and measuring the distances among them - a crude method that did not reliably result in matches. But recently, the technology has improved significantly, because of advances in artificial intelligence. A.I. software can analyze countless photos of people's faces and learn to make impressive predictions about which images are of the same person; the more faces it inspects, the better it gets. Clearview is deploying this approach using billions of photos from the public internet. By testing legal and ethical limits around the collection and use of those images, it has become the front-runner in the field. "
"The UK is not Russia. For all that the many civil liberty campaigners will complain, as is their role, the independence of the judiciary remains strong. The laws relating to freedom of association, expression and right to privacy are well defended in parliament and outside.
But the technology, the means by which the state might insert itself into our lives, is developing apace. The checks and balances are not. The Guardian has revealed that the government is legislating, without fanfare, to allow the police and the National Crime Agency to run facial recognition searches across the UK's driving licence records. When the police have an image, they will be able to identify the person, it is hoped, through the photographic images the state holds for the purposes of ensuring that the roads are safe.
Searching those digital images would have taken more man-hours than could have been justified in the old analogue world. It is now a matter of pushing a button, thanks to the wonders of artificial intelligence systems that are able to match biometric measurements in a flash."
"A leading psychologist who advises Meta on suicide prevention and self-harm has quit her role, accusing the tech giant of "turning a blind eye" to harmful content on Instagram, repeatedly ignoring expert advice and prioritising profit over lives.
Lotte Rubæk, who has been on Meta's global expert group for more than three years, told the Observer that the tech giant's ongoing failure to remove images of self-harm from its platforms is "triggering" vulnerable young women and girls to further harm themselves and contributing to rising suicide figures."
"If the NBA really cares about player and fan safety and racial justice, it is time that they ban facial recognition from their events and arenas. Facial recognition companies are forcefully marketing their racially biased products as false solutions to COVID-19-the NBA cannot buy these lies."